Generating Fair Universal Representations using Adversarial Models

نویسندگان

چکیده

We present a data-driven framework for learning fair universal representations (FUR) that guarantee statistical fairness any task may not be known priori. Our leverages recent advances in adversarial to allow data holder learn which set of sensitive attributes are decoupled from the rest dataset. formulate this as constrained minimax game between an encoder and adversary where constraint ensures measure usefulness (utility) representation. The resulting problem is censoring, i.e., finding representation least informative about given utility constraint. For appropriately chosen loss functions, our censoring precisely clarifies optimal strategy against strong information-theoretic adversaries; it also achieves demographic parity representations. evaluate performance proposed on both synthetic publicly available datasets. these datasets, we use two tradeoff measures: vs. fidelity downstream tasks, amply demonstrate multiple features can effectively censored even ensure accuracy tasks.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Learning Universal Adversarial Perturbations with Generative Models

Neural networks are known to be vulnerable to adversarial examples, inputs that have been intentionally perturbed to remain visually similar to the source input, but cause a misclassification. It was recently shown that given a dataset and classifier, there exists so called universal adversarial perturbations, a single perturbation that causes a misclassification when applied to any input. In t...

متن کامل

Generating models for temporal representations

We discuss the use of model building for temporal representations. We chose Polish to illustrate our discussion because it has an interesting aspectual system, but the points we wish to make are not language specific. Rather, our goal is to develop theoretical and computational tools for temporal model building tasks in computational semantics. To this end, we present a first-order theory of ti...

متن کامل

Combating Adversarial Attacks Using Sparse Representations

It is by now well-known that small adversarial perturbations can induce classification errors in deep neural networks (DNNs). In this paper, we make the case that sparse representations of the input data are a crucial tool for combating such attacks. For linear classifiers, we show that a sparsifying front end is provably effective against `∞-bounded attacks, reducing output distortion due to t...

متن کامل

Generating Adversarial Examples with Adversarial Networks

Deep neural networks (DNNs) have been found to be vulnerable to adversarial examples resulting from adding small-magnitude perturbations to inputs. Such adversarial examples can mislead DNNs to produce adversary-selected results. Different attack strategies have been proposed to generate adversarial examples, but how to produce them with high perceptual quality and more efficiently requires mor...

متن کامل

Generating Natural Adversarial Examples

Due to their complex nature, it is hard to characterize the ways in which machine learning models can misbehave or be exploited when deployed. Recent work on adversarial examples, i.e. inputs with minor perturbations that result in substantially different model predictions, is helpful in evaluating the robustness of these models by exposing the adversarial scenarios where they fail. However, th...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Information Forensics and Security

سال: 2022

ISSN: ['1556-6013', '1556-6021']

DOI: https://doi.org/10.1109/tifs.2022.3170265